From: Andres Lagar-Cavilla Date: Tue, 6 Dec 2011 21:16:56 +0000 (+0000) Subject: x86/mm: When mem event automatically promotes access rights, let other subsystems... X-Git-Url: https://dgit.raspbian.org/%22http:/www.example.com/cgi/%22https://%22%22/%22http:/www.example.com/cgi/%22https:/%22%22?a=commitdiff_plain;h=cddf900b88218075ae1b116cbe9937b621a9903a;p=xen.git x86/mm: When mem event automatically promotes access rights, let other subsystems know The mem event fault handler in the p2m can automatically promote the access rights of a p2m entry. In those scenarios, vcpu's are not paused and they will immediately retry the faulting instructions. This will generate a second fault if the underlying entry type requires so (paging, unsharing, pod, etc). Collapse the two faults into a single one. Signed-off-by: Andres Lagar-Cavilla Acked-by: Tim Deegan Committed-by: Tim Deegan --- diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 210e65abf3..0a63af88f2 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -1205,7 +1205,7 @@ int hvm_hap_nested_page_fault(unsigned long gpa, mfn_t mfn; struct vcpu *v = current; struct p2m_domain *p2m; - int rc; + int rc, fall_through = 0; /* On Nested Virtualization, walk the guest page table. * If this succeeds, all is fine. @@ -1278,9 +1278,15 @@ int hvm_hap_nested_page_fault(unsigned long gpa, if ( violation ) { - p2m_mem_access_check(gpa, gla_valid, gla, access_r, access_w, access_x); - rc = 1; - goto out_put_gfn; + if ( p2m_mem_access_check(gpa, gla_valid, gla, access_r, + access_w, access_x) ) + { + fall_through = 1; + } else { + /* Rights not promoted, vcpu paused, work here is done */ + rc = 1; + goto out_put_gfn; + } } } @@ -1339,7 +1345,11 @@ int hvm_hap_nested_page_fault(unsigned long gpa, goto out_put_gfn; } - rc = 0; + /* If we fell through, the vcpu will retry now that access restrictions have + * been removed. It may fault again if the p2m entry type still requires so. + * Otherwise, this is an error condition. */ + rc = fall_through; + out_put_gfn: put_gfn(p2m->domain, gfn); return rc; diff --git a/xen/arch/x86/mm/p2m.c b/xen/arch/x86/mm/p2m.c index 7363282179..fe93704549 100644 --- a/xen/arch/x86/mm/p2m.c +++ b/xen/arch/x86/mm/p2m.c @@ -1084,7 +1084,7 @@ void p2m_mem_paging_resume(struct domain *d) mem_event_unpause_vcpus(d); } -void p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla, +bool_t p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla, bool_t access_r, bool_t access_w, bool_t access_x) { struct vcpu *v = current; @@ -1105,7 +1105,7 @@ void p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla { p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2mt, p2m_access_rw); p2m_unlock(p2m); - return; + return 1; } p2m_unlock(p2m); @@ -1128,12 +1128,13 @@ void p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla p2m_lock(p2m); p2m->set_entry(p2m, gfn, mfn, PAGE_ORDER_4K, p2mt, p2m_access_rwx); p2m_unlock(p2m); + return 1; } - return; + return 0; } else if ( res > 0 ) - return; /* No space in buffer; VCPU paused */ + return 0; /* No space in buffer; VCPU paused */ memset(&req, 0, sizeof(req)); req.type = MEM_EVENT_TYPE_ACCESS; @@ -1157,6 +1158,7 @@ void p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla mem_event_put_request(d, &d->mem_event->access, &req); /* VCPU paused, mem event request sent */ + return 0; } void p2m_mem_access_resume(struct domain *d) diff --git a/xen/include/asm-x86/p2m.h b/xen/include/asm-x86/p2m.h index 8bb9142d5b..2e174bf184 100644 --- a/xen/include/asm-x86/p2m.h +++ b/xen/include/asm-x86/p2m.h @@ -491,8 +491,9 @@ static inline void p2m_mem_paging_populate(struct domain *d, unsigned long gfn) #ifdef __x86_64__ /* Send mem event based on the access (gla is -1ull if not available). Handles - * the rw2rx conversion */ -void p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla, + * the rw2rx conversion. Boolean return value indicates if access rights have + * been promoted with no underlying vcpu pause. */ +bool_t p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla, bool_t access_r, bool_t access_w, bool_t access_x); /* Resumes the running of the VCPU, restarting the last instruction */ void p2m_mem_access_resume(struct domain *d); @@ -508,10 +509,10 @@ int p2m_get_mem_access(struct domain *d, unsigned long pfn, hvmmem_access_t *access); #else -static inline void p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, +static inline bool_t p2m_mem_access_check(unsigned long gpa, bool_t gla_valid, unsigned long gla, bool_t access_r, bool_t access_w, bool_t access_x) -{ } +{ return 1; } static inline int p2m_set_mem_access(struct domain *d, unsigned long start_pfn, uint32_t nr, hvmmem_access_t access)